# Single GPU Training
Babyberta 3
MIT
BabyBERTa is a lightweight version based on RoBERTa, specifically designed for language acquisition research, trained on a 5-million-word corpus of American English child-directed input.
Large Language Model
Transformers English

B
phueb
22
0
Chinese Mobile Bert
Apache-2.0
This model was pre-trained on a 250-million-word Chinese corpus using the MobileBERT architecture, with a training period of 15 days, completing 1 million iterations on a single A100 GPU.
Large Language Model
Transformers

C
Ayou
25
5
Featured Recommended AI Models